17 research outputs found
The Foundations of Deep Learning with a Path Towards General Intelligence
Like any field of empirical science, AI may be approached axiomatically. We
formulate requirements for a general-purpose, human-level AI system in terms of
postulates. We review the methodology of deep learning, examining the explicit
and tacit assumptions in deep learning research. Deep Learning methodology
seeks to overcome limitations in traditional machine learning research as it
combines facets of model richness, generality, and practical applicability. The
methodology so far has produced outstanding results due to a productive synergy
of function approximation, under plausible assumptions of irreducibility and
the efficiency of back-propagation family of algorithms. We examine these
winning traits of deep learning, and also observe the various known failure
modes of deep learning. We conclude by giving recommendations on how to extend
deep learning methodology to cover the postulates of general-purpose AI
including modularity, and cognitive architecture. We also relate deep learning
to advances in theoretical neuroscience research.Comment: Submitted to AGI 201
Teraflop-scale Incremental Machine Learning
We propose a long-term memory design for artificial general intelligence
based on Solomonoff's incremental machine learning methods. We use R5RS Scheme
and its standard library with a few omissions as the reference machine. We
introduce a Levin Search variant based on Stochastic Context Free Grammar
together with four synergistic update algorithms that use the same grammar as a
guiding probability distribution of programs. The update algorithms include
adjusting production probabilities, re-using previous solutions, learning
programming idioms and discovery of frequent subprograms. Experiments with two
training sequences demonstrate that our approach to incremental learning is
effective
Gigamachine: incremental machine learning on desktop computers
We present a concrete design for Solomonoff's incremental machine learning
system suitable for desktop computers. We use R5RS Scheme and its standard
library with a few omissions as the reference machine. We introduce a Levin
Search variant based on a stochastic Context Free Grammar together with new
update algorithms that use the same grammar as a guiding probability
distribution for incremental machine learning. The updates include adjusting
production probabilities, re-using previous solutions, learning programming
idioms and discovery of frequent subprograms. The issues of extending the a
priori probability distribution and bootstrapping are discussed. We have
implemented a good portion of the proposed algorithms. Experiments with toy
problems show that the update algorithms work as expected.Comment: This is the original submission for my AGI-2010 paper titled
Stochastic Grammar Based Incremental Machine Learning Using Scheme which may
be found on http://agi-conf.org/2010/wp-content/uploads/2009/06/paper_24.pdf
and presented a partial but general solution to the transfer learning problem
in AI. arXiv admin note: substantial text overlap with arXiv:1103.100
Godseed: Benevolent or Malevolent?
It is hypothesized by some thinkers that benign looking AI objectives may
result in powerful AI drives that may pose an existential risk to human
society. We analyze this scenario and find the underlying assumptions to be
unlikely. We examine the alternative scenario of what happens when universal
goals that are not human-centric are used for designing AI agents. We follow a
design approach that tries to exclude malevolent motivations from AI agents,
however, we see that objectives that seem benevolent may pose significant risk.
We consider the following meta-rules: preserve and pervade life and culture,
maximize the number of free minds, maximize intelligence, maximize wisdom,
maximize energy production, behave like human, seek pleasure, accelerate
evolution, survive, maximize control, and maximize capital. We also discuss
various solution approaches for benevolent behavior including selfless goals,
hybrid designs, Darwinism, universal constraints, semi-autonomy, and
generalization of robot laws. A "prime directive" for AI may help in
formulating an encompassing constraint for avoiding malicious behavior. We
hypothesize that social instincts for autonomous robots may be effective such
as attachment learning. We mention multiple beneficial scenarios for an
advanced semi-autonomous AGI agent in the near future including space
exploration, automation of industries, state functions, and cities. We conclude
that a beneficial AI agent with intelligence beyond human-level is possible and
has many practical use cases
What Is It Like to Be a Brain Simulation?
We frame the question of what kind of subjective experience a brain
simulation would have in contrast to a biological brain. We discuss the brain
prosthesis thought experiment. We evaluate how the experience of the brain
simulation might differ from the biological, according to a number of
hypotheses about experience and the properties of simulation. Then, we identify
finer questions relating to the original inquiry, and answer them from both a
general physicalist, and panexperientialist perspective.Comment: 10 pages, draft of conference paper published in AGI 2012, also
accepted to AISB 2012 but it was too late to arrange travel, unfortunately;
Artificial General Intelligence, 5th International Conference, AGI 2012,
Oxford, UK, December 8-11, 2012. Proceeding
Ultimate Intelligence Part III: Measures of Intelligence, Perception and Intelligent Agents
We propose that operator induction serves as an adequate model of perception.
We explain how to reduce universal agent models to operator induction. We
propose a universal measure of operator induction fitness, and show how it can
be used in a reinforcement learning model and a homeostasis (self-preserving)
agent based on the free energy principle. We show that the action of the
homeostasis agent can be explained by the operator induction model.Comment: Third installation of the Ultimate Intelligence series. Submitted to
AGI-2017. arXiv admin note: text overlap with arXiv:1504.0330
1-D and 2-D Parallel Algorithms for All-Pairs Similarity Problem
All-pairs similarity problem asks to find all vector pairs in a set of
vectors the similarities of which surpass a given similarity threshold, and it
is a computational kernel in data mining and information retrieval for several
tasks. We investigate the parallelization of a recent fast sequential
algorithm. We propose effective 1-D and 2-D data distribution strategies that
preserve the essential optimizations in the fast algorithm. 1-D parallel
algorithms distribute either dimensions or vectors, whereas the 2-D parallel
algorithm distributes data both ways. Additional contributions to the 1-D
vertical distribution include a local pruning strategy to reduce the number of
candidates, a recursive pruning algorithm, and block processing to reduce
imbalance. The parallel algorithms were programmed in OCaml which affords much
convenience. Our experiments indicate that the performance depends on the
dataset, therefore a variety of parallelizations is useful
Diverse Consequences of Algorithmic Probability
We reminisce and discuss applications of algorithmic probability to a wide
range of problems in artificial intelligence, philosophy and technological
society. We propose that Solomonoff has effectively axiomatized the field of
artificial intelligence, therefore establishing it as a rigorous scientific
discipline. We also relate to our own work in incremental machine learning and
philosophy of complexity
Omega: An Architecture for AI Unification
We introduce the open-ended, modular, self-improving Omega AI unification
architecture which is a refinement of Solomonoff's Alpha architecture, as
considered from first principles. The architecture embodies several crucial
principles of general intelligence including diversity of representations,
diversity of data types, integrated memory, modularity, and higher-order
cognition. We retain the basic design of a fundamental algorithmic substrate
called an "AI kernel" for problem solving and basic cognitive functions like
memory, and a larger, modular architecture that re-uses the kernel in many
ways. Omega includes eight representation languages and six classes of neural
networks, which are briefly introduced. The architecture is intended to
initially address data science automation, hence it includes many problem
solving methods for statistical tasks. We review the broad software
architecture, higher-order cognition, self-improvement, modular neural
architectures, intelligent agents, the process and memory hierarchy, hardware
abstraction, peer-to-peer computing, and data abstraction facility.Comment: This is a high-level overview of the Omega AGI architecture which is
the basis of a data science automation system. Submitted to a worksho
Zeta Distribution and Transfer Learning Problem
We explore the relations between the zeta distribution and algorithmic
information theory via a new model of the transfer learning problem. The
program distribution is approximated by a zeta distribution with parameter near
. We model the training sequence as a stochastic process. We analyze the
upper temporal bound for learning a training sequence and its entropy rates,
assuming an oracle for the transfer learning problem. We argue from empirical
evidence that power-law models are suitable for natural processes. Four
sequence models are proposed. Random typing model is like no-free lunch where
transfer learning does not work. Zeta process independently samples programs
from the zeta distribution. A model of common sub-programs inspired by genetics
uses a database of sub-programs. An evolutionary zeta process samples mutations
from Zeta distribution. The analysis of stochastic processes inspired by
evolution suggest that AI may be feasible in nature, countering no-free lunch
sort of arguments.Comment: Submitted to AGI 2018, pre-prin